Knowledge-aware Leap-LSTM: Integrating Prior Knowledge into Leap-LSTM towards Faster Long Text Classification
نویسندگان
چکیده
While widely used in industry, recurrent neural networks (RNNs) are known to have deficiencies dealing with long sequences (e.g. slow inference, vanishing gradients etc.). Recent research has attempted accelerate RNN models by developing mechanisms skip irrelevant words input. Due the lack of labelled data, it remains as a challenge decide which skip, especially for low-resource classification tasks. In this paper, we propose Knowledge-AwareLeap-LSTM (KALL), novel architecture integrates prior human knowledge (created either manually or automatically) like in-domain keywords, terminologies lexicons into Leap-LSTM partially supervise skipping process. More specifically, knowledge-oriented cost function KALL; furthermore, two strategies integrate knowledge: (1) Factored KALL approach involves keyword indicator soft constraint skip-ping process, and (2) Gated enforces inclusion keywords while maintaining differentiable network training. Experiments on different public datasets show that our approaches are1.1x~2.6x faster than LSTM better accuracy 23.6x XLNet resource-limited CPU-only environment.
منابع مشابه
Integrating Background Knowledge Into Text Classification
We present a description of three different algorithms that use background knowledge to improve text classifiers. One uses the background knowledge as an index into the set of training examples. The second method uses background knowledge to reexpress the training examples. The last method treats pieces of background knowledge as unlabeled examples, and actually classifies them. The choice of b...
متن کاملLSTM-Based Mixture-of-Experts for Knowledge-Aware Dialogues
We introduce an LSTM-based method for dynamically integrating several wordprediction experts to obtain a conditional language model which can be good simultaneously at several subtasks. We illustrate this general approach with an application to dialogue where we integrate a neural chat model, good at conversational aspects, with a neural question-answering model, good at retrieving precise info...
متن کاملIntegrating Background Knowledge into Nearest-Neighbor Text Classification
This paper describes two different approaches for incorporating background knowledgeinto nearest-neighbor text classification.Our first approachuses backgroundtext to assessthe similarity betweentraining and test documentsrather than assessing their similarity directly. The second method redescribes examples using Latent Semantic Indexing on the background knowledge, assessing document similari...
متن کاملIntegrating Ontological Prior Knowledge into Relational Learning
Ontologies represent an important source of prior information which lends itself to the integration into statistical modeling. This paper discusses approaches towards employing ontological knowledge for relational learning. Our analysis is based on the IHRM model that performs relational learning by including latent variables that can be interpreted as cluster variables of the entities in the d...
متن کاملRobustly Leveraging Prior Knowledge in Text Classification
Prior knowledge has been shown very useful to address many natural language processing tasks. Many approaches have been proposed to formalise a variety of knowledge, however, whether the proposed approach is robust or sensitive to the knowledge supplied to the model has rarely been discussed. In this paper, we propose three regularization terms on top of generalized expectation criteria, and co...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2021
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v35i14.17511